Persistent Homology Lower Bounds on High-Order Network Distances
نویسندگان
چکیده
منابع مشابه
Tight Lower Bounds for Homology Inference
The homology groups of a manifold are important topological invariants that provide an algebraic summary of the manifold. These groups contain rich topological information, for instance, about the connected components, holes, tunnels and sometimes the dimension of the manifold. In earlier work [1], we have considered the statistical problem of estimating the homology of a manifold from noiseles...
متن کاملNew Lower Bounds on Subgroup Growth and Homology Growth
Subgroup growth is an important new area of group theory. It attempts to quantify the number of finite index subgroups of a group, as a function of their index. In this paper, we will provide new, strong lower bounds on the subgroup growth of a variety of different groups. This will include the fundamental groups of all finite-volume hyperbolic 3-manifolds. By using the correspondence between s...
متن کاملUsing persistent homology and dynamical distances to analyze protein binding.
Persistent homology captures the evolution of topological features of a model as a parameter changes. The most commonly used summary statistics of persistent homology are the barcode and the persistence diagram. Another summary statistic, the persistence landscape, was recently introduced by Bubenik. It is a functional summary, so it is easy to calculate sample means and variances, and it is st...
متن کاملLower bounds on two-terminal network reliability
Computing the probability that two nodes in a probabilistic network are connected is a wellknown computationally difficult problem. Two strategies are devised for obtaining lower bounds on the connection probability for two terminals. The first improves on the Kruskal-Katona bound by using efficient computations of small pathsets. The second strategy employs efficient algorithms for finding edg...
متن کاملLower Bounds for Higher-Order Convex Optimization
State-of-the-art methods in convex and non-convex optimization employ higher-order derivative information, either implicitly or explicitly. We explore the limitations of higher-order optimization and prove that even for convex optimization, a polynomial dependence on the approximation guarantee and higher-order smoothness parameters is necessary. As a special case, we show Nesterov’s accelerate...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Transactions on Signal Processing
سال: 2017
ISSN: 1053-587X,1941-0476
DOI: 10.1109/tsp.2016.2620963